OpenMP parallelism for fluid and fluid-particulate systems
نویسندگان
چکیده
0167-8191/$ see front matter 2012 Elsevier B.V http://dx.doi.org/10.1016/j.parco.2012.05.005 ⇑ Corresponding author. Tel.: +1 540 231 9975; fa E-mail address: [email protected] (D. Tafti). In order to exploit the flexibility of OpenMP in parallelizing large scale multi-physics applications where different modes of parallelism are needed for efficient computation, it is first necessary to be able to scale OpenMP codes as well as MPI on large core counts. In this research we have implemented fine grained OpenMP parallelism for a large CFD code GenIDLEST and investigated the performance from 1 to 256 cores using a variety of performance optimization and measurement tools. It is shown through weak and strong scaling studies that OpenMP performance can be made to match that of MPI on the SGI Altix systems for up to 256 cores. Data placement and locality were established to be key components in obtaining good scalability with OpenMP. It is also shown that a hybrid implementation on a dual core system gives the same performance as standalone MPI or OpenMP. Finally, it is shown that in irregular multi-physics applications which do not adhere solely to the SPMD (Single Process, Multiple Data) mode of computation, as encountered in tightly coupled fluid-particulate systems, the flexibility of OpenMP can have a big performance advantage over MPI. 2012 Elsevier B.V. All rights reserved.
منابع مشابه
A hybrid MPI-OpenMP scheme for scalable parallel pseudospectral computations for fluid turbulence
A hybrid scheme that utilizes MPI for distributed memory parallelism and OpenMP for shared memory parallelism is presented. The work is motivated by the desire to achieve exceptionally high Reynolds numbers in pseudospectral computations of fluid turbulence on emerging petascale, high core-count, massively parallel processing systems. The hybrid implementation derives from and augments a well-t...
متن کاملComparing the OpenMP, MPI, and Hybrid Programming Paradigms on an SMP Cluster
Clusters of SMP (Symmetric Multi-Processors) nodes provide support for a wide range of parallel programming paradigms. The shared address space within each node is suitable for OpenMP parallelization. Message passing can be employed within and across the nodes of a cluster. Multiple levels of parallelism can be achieved by combining message passing and OpenMP parallelization. Which programming ...
متن کاملAvailable on-line at www.prace-ri.eu Partnership for Advanced Computing in Europe OpenMP Parallelization of the Slilab Code
This white paper describes parallelization of the Slilab code with OpenMP for a shared-memory execution model when focusing on the multiphase phase flow simulations, such as fiber suspensions in turbulent channel flows. In such problems the motion of the ”second phase fibre” is frequently crossed over the distributed domain boundary of the ”first phase fluid”, which in turn reduces the work-bal...
متن کاملAutomatic Generation of OpenMP Directives and Its Application to Computational Fluid Dynamics Codes
The shared-memory programming model is a very effective way to achieve parallelism on shared memory parallel computers. As great progress was made in hardware and software technologies, performance of parallel programs with compiler directives has demonstrated large improvement. The introduction of OpenMP directives, the industrial standard for shared-memory programming, has minimized the issue...
متن کاملPerformance Evaluation of Numeric Compute Kernels on nVIDIA GPUs
Graphics processing units provide an astonishing number of floating point operations per second and deliver memory bandwidths of one magnitude greater than common general purpose central processing units. With the introduction of the Compute Unified Device Architecture, a first step was taken by nVIDIA to ease access to the vast computational resources of graphics processing units. The aim of t...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
- Parallel Computing
دوره 38 شماره
صفحات -
تاریخ انتشار 2012